US Virgin Islands
Jeffrey Epstein's Ties to CBP Agents Sparked a DOJ Probe
Documents say customs officers in the US Virgin Islands had friendly relationships with Epstein years after his 2008 conviction, showing how the infamous sex offender tried to cultivate allies. United States prosecutors and federal law enforcement spent over a year examining ties between Jeffrey Epstein and Customs and Border Protection officers stationed in the US Virgin Islands (USVI), according to documents recently released by the Department of Justice. As The Guardian and New York Times have reported, emails, text messages, and investigative records show that Epstein cultivated friendships with several officers, entertaining them on his island and offering to take them for whale-watching trips in his helicopter. He even brought one cannolis for Christmas Eve. In turn, Epstein would bring certain officers his complaints about his treatment at the hands of other CBP and federal agents.
- North America > US Virgin Islands (0.81)
- North America > United States > California (0.14)
- North America > United States > New York (0.05)
- (3 more...)
- Law Enforcement & Public Safety > Crime Prevention & Enforcement (1.00)
- Law (1.00)
- Government > Regional Government > North America Government > United States Government (1.00)
- Government > Immigration & Customs (1.00)
- Information Technology > Artificial Intelligence (0.47)
- Information Technology > Communications > Mobile (0.35)
- Asia > Japan > Honshū > Tōhoku > Iwate Prefecture > Morioka (0.04)
- North America > United States > California (0.04)
- Asia > China > Guangdong Province > Shenzhen (0.04)
- (3 more...)
- Research Report > Experimental Study (0.93)
- Research Report > New Finding (0.67)
- Energy > Power Industry (1.00)
- Energy > Renewable > Solar (0.93)
Evaluation of A Spatial Microsimulation Framework for Small-Area Estimation of Population Health Outcomes Using the Behavioral Risk Factor Surveillance System
Von Hoene, Emma, Gupta, Aanya, Kavak, Hamdi, Roess, Amira, Anderson, Taylor
The field of population health addresses a wide spectrum of challenges, spanning infectious and chronic diseases to mental health and health risk behaviors such as smoking and alcohol consumption (Sharma et al., 2025). A common barrie r to addressing these issues is the lack of ground truth data capturing health outcomes and behaviors at fine geographic scales. This limits both local and national health decision - makers in planning and management efforts, such as identify ing health inequalities or targeting interventions where they are most needed (Rahman, 2017; Wang, 2018) . T o fill this gap, researchers use small area estimation (SAE), a collection of statistical methods that combine survey and geographic data to generate estimates of population - level health outcomes at various spatial scales (RTI International, 2025) . There are numerous methods for generating SAE of health outcomes, which can generally be grouped into two main approaches: direct and indirect model - based estimates (Rahman, 2017) . Direct estimates are calculated using only the survey responses from individuals or households sampled within the specified geographi c areas (counties, states) to estimate disease prevalence or other population characteristics.
- North America > United States > New York (0.05)
- North America > United States > Virginia > Fairfax County > Fairfax (0.04)
- South America > Chile > Santiago Metropolitan Region > Santiago Province > Santiago (0.04)
- (11 more...)
- Questionnaire & Opinion Survey (1.00)
- Research Report > New Finding (0.93)
- Research Report > Experimental Study (0.67)
- Asia > Japan > Honshū > Tōhoku > Iwate Prefecture > Morioka (0.04)
- North America > United States > California (0.04)
- Asia > China > Guangdong Province > Shenzhen (0.04)
- (3 more...)
- Research Report > Experimental Study (0.93)
- Research Report > New Finding (0.67)
- Energy > Power Industry (1.00)
- Energy > Renewable > Solar (0.93)
Predicting Microbial Ontology and Pathogen Risk from Environmental Metadata with Large Language Models
Traditional machine learning models struggle to generalize in microbiome studies where only metadata is available, especially in small-sample settings or across studies with heterogeneous label formats. In this work, we explore the use of large language models (LLMs) to classify microbial samples into ontology categories such as EMPO 3 and related biological labels, as well as to predict pathogen contamination risk, specifically the presence of E. Coli, using environmental metadata alone. We evaluate LLMs such as ChatGPT-4o, Claude 3.7 Sonnet, Grok-3, and LLaMA 4 in zero-shot and few-shot settings, comparing their performance against traditional models like Random Forests across multiple real-world datasets. Our results show that LLMs not only outperform baselines in ontology classification, but also demonstrate strong predictive ability for contamination risk, generalizing across sites and metadata distributions. These findings suggest that LLMs can effectively reason over sparse, heterogeneous biological metadata and offer a promising metadata-only approach for environmental microbiology and biosurveillance applications.
- North America > US Virgin Islands (0.09)
- North America > Aruba (0.05)
- North America > United States > Ohio (0.04)
- (3 more...)
Deep learning four decades of human migration
W e present a novel and detailed dataset on origin-destination annual migration flows and stocks between 230 countries and regions, spanning the period from 1990 to the present. Our flow estimates are further disaggregated by country of birth, providing a comprehensive picture of migration over the last 35 years. The estimates are obtained by training a deep recurrent neural network to learn flow patterns from 18 covariates for all countries, including geographic, economic, cultural, societal, and political information. The recurrent architecture of the neural network means that the entire past can influence current migration patterns, allowing us to learn long-range temporal correlations. By training an ensemble of neural networks and additionally pushing uncertainty on the covariates through the trained network, we obtain confidence bounds for all our estimates, allowing researchers to pinpoint the geographic regions most in need of additional data collection. W e validate our approach on various test sets of unseen data, demonstrating that it significantly outperforms traditional methods estimating five-year flows while delivering a significant increase in temporal resolution. The model is fully open source: all training data, neural network weights, and training code are made public alongside the migration estimates, providing a valuable resource for future studies of human migration.
- Oceania > Australia (0.46)
- Europe > Isle of Man (0.28)
- Asia > Russia (0.28)
- (94 more...)
Simulating Correlated Electrons with Symmetry-Enforced Normalizing Flows
Schuh, Dominic, Kreit, Janik, Berkowitz, Evan, Funcke, Lena, Luu, Thomas, Nicoli, Kim A., Rodekamp, Marcel
One of the most widely used theoretical frameworks for studying such systems is the Hubbard model [2-4], which captures the essential competition between electron kinetic energy and on-site interactions. Over the years, a variety of methods have been developed to analyze the Hubbard model. In the weak interaction regime, perturbative approaches provide valuable insights [5]. However, outside this regime, non-perturbative effects become significant, rendering pertur-bative techniques insufficient. In these regimes, Monte Carlo simulations become an indispensable tool (see, for example, [6-13] and references therein).
- Europe > Germany > North Rhine-Westphalia > Cologne Region > Bonn (0.04)
- North America > US Virgin Islands (0.04)
- Europe > Germany > Bavaria > Regensburg (0.04)
Revisiting Noise in Natural Language Processing for Computational Social Science
Computational Social Science (CSS) is an emerging field driven by the unprecedented availability of human-generated content for researchers. This field, however, presents a unique set of challenges due to the nature of the theories and datasets it explores, including highly subjective tasks and complex, unstructured textual corpora. Among these challenges, one of the less well-studied topics is the pervasive presence of noise. This thesis aims to address this gap in the literature by presenting a series of interconnected case studies that examine different manifestations of noise in CSS. These include character-level errors following the OCR processing of historical records, archaic language, inconsistencies in annotations for subjective and ambiguous tasks, and even noise and biases introduced by large language models during content generation. This thesis challenges the conventional notion that noise in CSS is inherently harmful or useless. Rather, it argues that certain forms of noise can encode meaningful information that is invaluable for advancing CSS research, such as the unique communication styles of individuals or the culture-dependent nature of datasets and tasks. Further, this thesis highlights the importance of nuance in dealing with noise and the considerations CSS researchers must address when encountering it, demonstrating that different types of noise require distinct strategies.
- North America > United States > Minnesota > Hennepin County > Minneapolis (0.14)
- Europe > Poland (0.14)
- Europe > Finland (0.14)
- (130 more...)
- Research Report > New Finding (1.00)
- Research Report > Experimental Study (1.00)
- Questionnaire & Opinion Survey (1.00)
- (2 more...)
- Media > News (1.00)
- Leisure & Entertainment (1.00)
- Law Enforcement & Public Safety > Crime Prevention & Enforcement (1.00)
- (10 more...)
Simulating the Hubbard Model with Equivariant Normalizing Flows
Schuh, Dominic, Kreit, Janik, Berkowitz, Evan, Funcke, Lena, Luu, Thomas, Nicoli, Kim A., Rodekamp, Marcel
Generative models, particularly normalizing flows, have shown exceptional performance in learning probability distributions across various domains of physics, including statistical mechanics, collider physics, and lattice field theory. In the context of lattice field theory, normalizing flows have been successfully applied to accurately learn the Boltzmann distribution, enabling a range of tasks such as direct estimation of thermodynamic observables and sampling independent and identically distributed (i.i.d.) configurations. In this work, we present a proof-of-concept demonstration that normalizing flows can be used to learn the Boltzmann distribution for the Hubbard model. This model is widely employed to study the electronic structure of graphene and other carbon nanomaterials. State-of-the-art numerical simulations of the Hubbard model, such as those based on Hybrid Monte Carlo (HMC) methods, often suffer from ergodicity issues, potentially leading to biased estimates of physical observables. Our numerical experiments demonstrate that leveraging i.i.d.\ sampling from the normalizing flow effectively addresses these issues.
- Europe > Germany > North Rhine-Westphalia > Cologne Region > Bonn (0.04)
- Europe > Germany > Bavaria > Regensburg (0.04)
- North America > US Virgin Islands (0.04)
- Europe > United Kingdom > England > Merseyside > Liverpool (0.04)
Establishing Nationwide Power System Vulnerability Index across US Counties Using Interpretable Machine Learning
Ma, Junwei, Li, Bo, Omitaomu, Olufemi A., Mostafavi, Ali
Power outages have become increasingly frequent, intense, and prolonged in the US due to climate change, aging electrical grids, and rising energy demand. However, largely due to the absence of granular spatiotemporal outage data, we lack data-driven evidence and analytics-based metrics to quantify power system vulnerability. This limitation has hindered the ability to effectively evaluate and address vulnerability to power outages in US communities. Here, we collected ~179 million power outage records at 15-minute intervals across 3022 US contiguous counties (96.15% of the area) from 2014 to 2023. We developed a power system vulnerability assessment framework based on three dimensions (intensity, frequency, and duration) and applied interpretable machine learning models (XGBoost and SHAP) to compute Power System Vulnerability Index (PSVI) at the county level. Our analysis reveals a consistent increase in power system vulnerability over the past decade. We identified 318 counties across 45 states as hotspots for high power system vulnerability, particularly in the West Coast (California and Washington), the East Coast (Florida and the Northeast area), the Great Lakes megalopolis (Chicago-Detroit metropolitan areas), and the Gulf of Mexico (Texas). Heterogeneity analysis indicates that urban counties, counties with interconnected grids, and states with high solar generation exhibit significantly higher vulnerability. Our results highlight the significance of the proposed PSVI for evaluating the vulnerability of communities to power outages. The findings underscore the widespread and pervasive impact of power outages across the country and offer crucial insights to support infrastructure operators, policymakers, and emergency managers in formulating policies and programs aimed at enhancing the resilience of the US power infrastructure.
- North America > Mexico (0.34)
- North America > United States > Illinois > Cook County > Chicago (0.24)
- Atlantic Ocean > Gulf of Mexico (0.24)
- (41 more...)
- Machinery > Industrial Machinery (1.00)
- Government > Regional Government > North America Government > United States Government (1.00)
- Energy > Renewable > Wind (1.00)
- (2 more...)